29 research outputs found
Adaptive Path Planning for Depth Constrained Bathymetric Mapping with an Autonomous Surface Vessel
This paper describes the design, implementation and testing of a suite of
algorithms to enable depth constrained autonomous bathymetric (underwater
topography) mapping by an Autonomous Surface Vessel (ASV). Given a target depth
and a bounding polygon, the ASV will find and follow the intersection of the
bounding polygon and the depth contour as modeled online with a Gaussian
Process (GP). This intersection, once mapped, will then be used as a boundary
within which a path will be planned for coverage to build a map of the
Bathymetry. Methods for sequential updates to GP's are described allowing
online fitting, prediction and hyper-parameter optimisation on a small embedded
PC. New algorithms are introduced for the partitioning of convex polygons to
allow efficient path planning for coverage. These algorithms are tested both in
simulation and in the field with a small twin hull differential thrust vessel
built for the task.Comment: 21 pages, 9 Figures, 1 Table. Submitted to The Journal of Field
Robotic
Machine-learning of atomic-scale properties based on physical principles
We briefly summarize the kernel regression approach, as used recently in
materials modelling, to fitting functions, particularly potential energy
surfaces, and highlight how the linear algebra framework can be used to both
predict and train from linear functionals of the potential energy, such as the
total energy and atomic forces. We then give a detailed account of the Smooth
Overlap of Atomic Positions (SOAP) representation and kernel, showing how it
arises from an abstract representation of smooth atomic densities, and how it
is related to several popular density-based representations of atomic
structure. We also discuss recent generalisations that allow fine control of
correlations between different atomic species, prediction and fitting of
tensorial properties, and also how to construct structural kernels---applicable
to comparing entire molecules or periodic systems---that go beyond an additive
combination of local environments
Incremental Gaussian Processes
In this paper, we consider Tipping鈥榮 relevance vector machine (RVM) and formalize an incremental training strategy as a variant of the expectation-maximization (EM) algorithm that we call subspace EM. Working with a subset of active basis functions, the sparsity of the RVM solution will ensure that the number of basis functions and thereby the computational complexity is kept low. We also introduce a mean field approach to the intractable classification model that is expected to give a very good approximation to exact Bayesian inference and contains the Laplace approximation as a special case. We test the algorithms on two large data sets with O(10^3-10^4) examples. The results indicate that Bayesian learning of large data sets, e.g. the MNIST database is realistic
A Unifying View of Sparse Approximate Gaussian Process Regression
We provide a new unifying view, including all existing proper probabilistic sparse approximations for Gaussian process regression. Our approach relies on expressing the effective prior which the methods are using. This allows new insights to be gained, and highlights the relationship between existing methods. It also allows for a clear theoretically justified ranking of the closeness of the known approximations to the corresponding full GPs. Finally we point directly to designs of new better sparse approximations, combining the best of the existing strategies, within attractive computational constraints
Machine Learning Challenges. Evaluating Predictive Uncertainty, Visual Object Classification, and Recognising Tectual Entailment: First PASCAL Machine Learning Challenges Workshop, MLCW 2005, Southampton, UK, April 11-13, 2005
This book constitutes the thoroughly refereed post-proceedings of the First PASCAL (pattern analysis, statistical modelling and computational learning) Machine Learning Challenges Workshop, MLCW 2005, held in Southampton, UK in April 2005. The 25 revised full papers presented were carefully selected during two rounds of reviewing and improvement from about 50 submissions. The papers reflect the concepts of three challenges dealt with in the workshop: finding an assessment base on the uncertainty of predictions using classical statistics, Bayesian inference, and statistical learning theory; the second challenge was to recognize objects from a number of visual object classes in realistic scenes; the third challenge of recognizing textual entailment addresses semantic analysis of language to form a generic framework for applied semantic inference in text understanding
Sparse Spectrum Gaussian Process Regression
We present a new sparse Gaussian Process (GP) model for regression. The key novel idea is to sparsify the spectral representation of the GP. This leads to a simple, practical algorithm for regression tasks. We compare the achievable trade-offs between predictive accuracy and computational requirements, and show that these are typically superior to existing state-of-the-art sparse approximations. We discuss both the weight space and function space representations, and note that the new construction implies priors over functions which are always stationary, and can approximate any covariance function in this class